Enhancing multimodal silent speech interfaces with feature selection
نویسندگان
چکیده
In research on Silent Speech Interfaces (SSI), different sources of information (modalities) have been combined, aiming at obtaining better performance than the individual modalities. However, when combining these modalities, the dimensionality of the feature space rapidly increases, yielding the well-known “curse of dimensionality”. As a consequence, in order to extract useful information from this data, one has to resort to feature selection (FS) techniques to lower the dimensionality of the learning space. In this paper, we assess the impact of FS techniques for silent speech data, in a dataset with 4 non-invasive and promising modalities, namely: video, depth, ultrasonic Doppler sensing, and surface electromyography. We consider two supervised (mutual information and Fisher’s ratio) and two unsupervised (meanmedian and arithmetic mean geometric mean) FS filters. The evaluation was made by assessing the classification accuracy (word recognition error) of three well-known classifiers (knearest neighbors, support vector machines, and dynamic time warping). The key results of this study show that both unsupervised and supervised FS techniques improve on the classification accuracy on both individual and combined modalities. For instance, on the video component, we attain relative performance gains of 36.2% in error rates. FS is also useful as pre-processing for feature fusion.
منابع مشابه
Context-aware Multimodal Interfaces Enhancing Ubiquitous Learning
This paper proposes multimodal interfaces for a mobile learning system called SCROLL (System for Capturing and Reminding of Learning Log). The system is designed to help learners to record what they have learned (called as learning log) and recall their learning logs by providing them with quizzes. With the help of multimodal interfaces, learners can interact with the system by speech and body ...
متن کاملMultimodal Silent Speech Interface based on Video, Depth, Surface Electromyography and Ultrasonic Doppler: Data Collection and First Recognition Results
Silent Speech Interfaces use data from the speech production process, such as visual information of face movements. However, using a single modality limits the amount of available information. In this study we start to explore the use of multiple data input modalities in order to acquire a more complete representation of the speech production model. We have selected 4 non-invasive modalities – ...
متن کاملEnabling Multimodal Mobile Interfaces for Musical Performance
We present research that extends the scope of the mobile application Control, a prototyping environment for defining multimodal interfaces that control real-time artistic and musical performances. Control allows users to rapidly create interfaces employing a variety of modalities, including: speech recognition, computer vision, musical feature extraction, touchscreen widgets, and inertial senso...
متن کاملImproving of Feature Selection in Speech Emotion Recognition Based-on Hybrid Evolutionary Algorithms
One of the important issues in speech emotion recognizing is selecting of appropriate feature sets in order to improve the detection rate and classification accuracy. In last studies researchers tried to select the appropriate features for classification by using the selecting and reducing the space of features methods, such as the Fisher and PCA. In this research, a hybrid evolutionary algorit...
متن کاملEnabling Multimodal Mobile Interfaces for Interactive Musical Performance
We present research that extends the scope of the mobile application Control, a prototyping environment for defining multimodal interfaces that control real-time artistic and musical performances. Control allows users to rapidly create interfaces employing a variety of modalities, including: speech recognition, computer vision, musical feature extraction, touchscreen widgets, and inertial senso...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2014